Online learning from nite training sets
نویسندگان
چکیده
We analyse online gradient descent learning from nite training sets at non-innnitesimal learning rates for both linear and non-linear networks. In the linear case, exact results are obtained for the time-dependent generalization error of networks with a large number of weights N, trained on p = N examples. This allows us to study in detail the eeects of nite training set size on, for example, the optimal choice of learning rate. We also compare online and oine learning , for respective optimal settings of at given nal learning time. Online learning turns out to be much more robust to input bias and actually outperforms oine learning when such bias is present; for un-biased inputs, online and oine learning perform almost equally well. Our analysis of online learning for non-linear networks (namely, soft-committee machines), advances the theory to more realistic learning scenarios. Dynamical equations are derived for an appropriate set of order parameters; these are exact in the limiting case of either linear networks or innnite training sets. Preliminary comparisons with simulations suggest that the theory captures some eects of nite training sets, but may not yet account correctly for the presence of local minima .
منابع مشابه
Online learning from nite training sets androbustness to input
We analyse online gradient descent learning from nite training sets at non-innnitesimal learning rates. Exact results are obtained for the time-dependent generalization error of a simple model system: a linear network with a large number of weights N, trained on p = N examples. This allows us to study in detail the eeects of nite training set size on, for example, the optimal choice of learning...
متن کاملLearning from Finite Training Sets
{ We analyse online (gradient descent) learning of a rule from a nite set of training examples at non-innnitesimal learning rates , calculating exactly the time-dependent generalization error for a simple model scenario. In the thermodynamic limit, we close the dynamical equation for the generating function of an innnite hierarchy of order parameters using`within-sample self-averaging'. The res...
متن کاملOn-line Learning from Finite Training Sets in Nonlinear Networks
Online learning is one of the most common forms of neural network training. We present an analysis of online learning from finite training sets for non-linear networks (namely, soft-committee machines), advancing the theory to more realistic learning scenarios. Dynamical equations are derived for an appropriate set of order parameters; these are exact in the limiting case of either linear netwo...
متن کاملOnline Learning from Finite Training Sets: An Analytical Case Study
We analyse online learning from finite training sets at noninfinitesimal learning rates TJ. By an extension of statistical mechanics methods, we obtain exact results for the time-dependent generalization error of a linear network with a large number of weights N. We find, for example, that for small training sets of size p ~ N, larger learning rates can be used without compromising asymptotic g...
متن کاملCoronavirus and medical education: Getting change and turning to new education paradigm: the need for scaffolding learning in online training
Coronavirus pandemic has caused extensive changes in various fields. Education is one challenging and necessary dimension to be paid attention by policymakers and considered as a necessity. Due to changes in the methods and styles of providing medical education from face-to-face to virtual training, and due to its novelty in our country and several challenges ahead, it is necessary to take meas...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1998